Using User Models in Music Information Retrieval Systems
نویسندگان
چکیده
Most websites providing music services only support category-based browsing and/or text-based searching. There has been some research to improve the interface either for pull applications, e.g. query-by-humming systems, or for push applications, e.g. collaborative-filtering-based or featurebased music recommendation systems. However, for content-based search or feature-based filtering systems, one important problem is to describe music by its parameters or features, so that search engines or information filtering agents can use them to measure the similarity of the target (userÕs query or preference) and the candidates. MPEG7 (formally called ÒMultimedia Content Description InterfaceÓ) is an international standard, which describes the multimedia content data to allow universal indexing, retrieval, filtering, control, and other activities supported by rich metadata. However, the metadata about the multimedia content itself are still insufficient, because many features of multimedia content are quite perceptual and user-dependent. For example, emotional features are very important for multimedia retrieval, but they are hard to be described by a universal model since different users may have different emotional responses to the same multimedia content. We therefore turn to user modeling techniques and representations to describe the properties of each user, so that the retrieval will be more accurate. Besides, user modeling can be used to reduce the search space, make push service easier and improve the user interface. There are several important issues in user modeling for music information retrieval purpose or even more general multimedia retrieval. 1) How to model the user? User-programmed, machinelearning and knowledge-engineered methods can be used. 2) What information is needed to describe a user for music IR purpose? It may include both the userÕs indirect information (e.g. age, sex, citizenship, education, music experience, etc.) and direct information (e.g. userÕs interests, definition of qualitative features, appreciation habit, etc.). 3) How to represent, use and share the user model? Similar to MPEG7 concepts, we can use a standard language in text format to represent the user model, so that search engines or information filtering agents can use it to refine the result easily and efficiently, without repeating the long-time observation and learning of the userÕs behavior. User modeling can be done on client-side or server-side. Issues including easy/hard to obtain the user information, hard/easy to use collaborative filtering model, far from/close to the music data, more/less privacy or safety, more scalable/higher load on the server, etc., need to be considered when choosing either of the paradigms. We adopted the client-side user modeling paradigm in our MusicCat system. It is an agent that allows the user to define contexts and corresponding features of music that he wants to hear in those contexts correspondingly. Besides, the user can also define qualitative features of music based on quantitative features. For examples, the user just needs to tell the agent what kind of music he prefers to hear at what kind of context, like ÒI need fast and exciting music when IÕm happyÓ, ÒI need soft music to wake me up every morning at 8:00Ó, ÒI need slow classical music, when IÕm thinkingÓ, ÒI need rhythmic music when IÕm walkingÓ, etc. Or, the user can define qualitative features, like ÒRomantic music for me means slow music with titles or lyric including word loveÓ, ÒMy favorite music includes ÉÓ etc. Then, when the moment comes the user tells the agent or a pre-defined time approaches, the agent can automatically, randomly and repeatedly choose music from the userÕs collection according to the pre-defined constraints. So far, MusicCat uses a profile-based user interface. And only midi files are used in the system. In our system, we categorize the quantitative features of music into textual features including title, composer, genre, country, etc.; notation features including key signature, time signature, tonality, BPM, etc.; perceptual features including slow or fast, short or long, quiet or loud, nonrhythmic or rhythmic, and soft or exciting. User can define music corresponding to a particular context or qualitative feature based on these quantitative features. We manually input the textual information. Notation features were automatically extracted from midi files. Perceptual features were computed from the parameters including average duration per note, average tempo, tempo deviation, average pitch change per note, etc. The weights of each parameter contributing to each perceptual feature were set manually. Some psychoacoustic experiments should be necessary for more accuracy. To make the system more appealing, wireless communication technology can be used to make the system portable. We may also use sensors that can automatically detect the user context and then play corresponding music. User information is very valuable and needs to be shared in the future to make universal information retrieval possible. Thus we propose an XML-like language, UMIRL (User Modeling for Information Retrieval Language), in which different systems may describe the user in this standard format to make the user model sharable and reusable. Although it is designed for music information retrieval purpose in this paper, it can be extended for general multimedia information retrieval as well. To make music information retrieval systems more efficient, both user modeling techniques and descriptions are important. Those are prerequisites for open and efficient personalized services. We propose the user modeling language because there hasnÕt been much attention on the ways to share the valuable user data. We believe the main issues discussed in this paper will be the most significant but also the hardiest points in this research area.
منابع مشابه
Prototyping a Vibrato-Aware Query-By-Humming (QBH) Music Information Retrieval System for Mobile Communication Devices: Case of Chromatic Harmonica
Background and Aim: The current research aims at prototyping query-by-humming music information retrieval systems for smart phones. Methods: This multi-method research follows simulation technique from mixed models of the operations research methodology, and the documentary research method, simultaneously. Two chromatic harmonica albums comprised the research population. To achieve the purpose ...
متن کاملAnalaysis of IFLA Library Refrence Model’s Entities and Attrbutes For Iranian Traditional Music Resources (Case study: Morq-e sahar Song)
Background and Aim: The object of the study was to Analyze IFLA Library Reference Model (LRM) Entities and Attributes for Iranian Traditional Music Resources, Case Study: Morq-e Sahar Song. Method: The study inherits an applied content analysis method. All Entities and Attributes of IFlA LRM Model based on two checklists include: Final report of IFlA LRM on August 2017 and Transition Mappi...
متن کاملBehavioral Considerations in Developing Web Information Systems: User-centered Design Agenda
The current paper explores designing a web information retrieval system regarding the searching behavior of users in real and everyday life. Designing an information system that is closely linked to human behavior is equally important for providers and the end users. From an Information Science point of view, four approaches in designing information retrieval systems were identified as system-...
متن کاملEvaluation of Music Information Retrieval: Towards a User-Centered Approach
With the dramatic increase of online digital music, research on Music Information Retrieval (MIR) is flourishing more than ever. However, evaluation in MIR has been focused on system-centered approaches, where systems are evaluated against a pre-built ground truth dataset using system-focused measurements, and little attention has been spent on user experience. In this paper, we argue that MIR ...
متن کاملUser-centered Measures vs. System Effectiveness in Finding Similar Songs
User evaluation in the domain of Music Information Retrieval (MIR) has been very scarce, while algorithms and systems in MIR have been improving rapidly. With the maturity of system-centered evaluation in MIR, time is ripe for MIR evaluation to involve users. In this study, we compare user-centered measures to a system effectiveness measure on the task of retrieving similar songs. To collect us...
متن کامل